翻訳と辞書
Words near each other
・ Dirichlet algebra
・ Dirichlet beta function
・ Dirichlet boundary condition
・ Dirichlet character
・ Dirichlet conditions
・ Dirichlet convolution
・ Dirichlet density
・ Dirichlet distribution
・ Dirichlet eigenvalue
・ Dirichlet eta function
・ Dirichlet form
・ Dirichlet integral
・ Dirichlet kernel
・ Dirichlet L-function
・ Dirichlet problem
Dirichlet process
・ Dirichlet series
・ Dirichlet space
・ Dirichlet's approximation theorem
・ Dirichlet's energy
・ Dirichlet's principle
・ Dirichlet's test
・ Dirichlet's theorem
・ Dirichlet's theorem on arithmetic progressions
・ Dirichlet's unit theorem
・ Dirichlet-multinomial distribution
・ Dirichletia
・ Dirick Carver
・ Dirico
・ Dirico Airport


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

Dirichlet process : ウィキペディア英語版
Dirichlet process

In probability theory, Dirichlet processes (after Peter Gustav Lejeune Dirichlet) are a family of stochastic processes whose realizations are probability distributions. In other words, a Dirichlet process is a probability distribution whose domain is itself a set of probability distributions. It is often used in Bayesian inference to describe the prior knowledge about the distribution of random variables, that is, how likely it is that the random variables are distributed according to one or another particular distribution.
The Dirichlet process is specified by a base distribution H and a positive real number \alpha called the concentration parameter. The base distribution is the expected value of the process, that is, the Dirichlet process draws distributions "around" the base distribution in the way that a normal distribution draws real numbers around its mean. However, even if the base distribution is continuous, the distributions drawn from the Dirichlet process are almost surely discrete. The concentration parameter specifies how strong this discretization is: in the limit of \alpha\rightarrow 0, the realizations are all concentrated on a single value, while in the limit of \alpha\rightarrow\infty the realizations become continuous. In between the two extremes the realizations are discrete distributions with less and less concentration as \alpha increases.
The Dirichlet process can also be seen as the infinite-dimensional generalization of the Dirichlet distribution. In the same way as the Dirichlet distribution is the conjugate prior for the categorical distribution, the Dirichlet process is the conjugate prior for infinite, nonparametric discrete distributions. A particularly important application of Dirichlet processes is as a prior probability distribution in infinite mixture models.
The Dirichlet process was formally introduced by Thomas Ferguson in 1973
and has since been applied in data mining and machine learning, among others for natural language processing, computer vision and bioinformatics.
==Introduction==
Dirichlet processes are usually used when modeling data that tends to repeat previous values in a "rich get richer" fashion. Specifically, suppose that the generation of values X_,X_,\dots can be simulated by the following algorithm.
:Input: H (a probability distribution called base distribution), \alpha (a positive real number called concentration parameter)
# Draw X_ from the distribution H.
# For n>1:
## With probability \frac draw X_ from H.
## With probability \frac set X_=x, where n_ is the number of previous observations X_, j, such that X_=x.
At the same time, another common model for data is that the observations X_,X_,\dots are assumed to be independent and identically distributed (i.i.d.) according to some distribution P. The goal in introducing Dirichlet processes is to be able to describe the procedure outlined above in this i.i.d. model.
The X_,X_,\dots observations are not independent, since we have to consider the previous results when generating the next value. They are, however, exchangeable. This fact can be shown by calculating the joint probability distribution of the observations and noticing that the resulting formula only depends on which x values occur among the observations and how many repetitions they each have. Because of this exchangeability, de Finetti's representation theorem applies and it implies that the observations X_,X_,\dots are conditionally independent given a (latent) distribution P. This P is a random variable itself and has a distribution. This distribution (over distributions) is called Dirichlet process (\mathrm). In summary, this means that we get an equivalent procedure to the above algorithm:
# Draw a distribution P from \mathrm\left(H,\alpha\right)
# Draw observations X_,X_\dots independently from P.
In practice, however, drawing a concrete distribution P is impossible, since its specification requires an infinite amount of information. This is a common phenomenon in the context of Bayesian non-parametric statistics where a typical task is to learn distributions on function spaces, which involve effectively infinitely many parameters. The key insight is that in many applications the infinite dimensional distributions appear only as an intermediary computational device and are not required for either the initial specification of prior beliefs or for the statement of the final inference. The Dirichlet process can be used to circumvent infinite computational requirements as described above.

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「Dirichlet process」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.